Optimization of tomographic reconstruction workflows on geographically distributed resources.
نویسندگان
چکیده
New technological advancements in synchrotron light sources enable data acquisitions at unprecedented levels. This emergent trend affects not only the size of the generated data but also the need for larger computational resources. Although beamline scientists and users have access to local computational resources, these are typically limited and can result in extended execution times. Applications that are based on iterative processing as in tomographic reconstruction methods require high-performance compute clusters for timely analysis of data. Here, time-sensitive analysis and processing of Advanced Photon Source data on geographically distributed resources are focused on. Two main challenges are considered: (i) modeling of the performance of tomographic reconstruction workflows and (ii) transparent execution of these workflows on distributed resources. For the former, three main stages are considered: (i) data transfer between storage and computational resources, (i) wait/queue time of reconstruction jobs at compute resources, and (iii) computation of reconstruction tasks. These performance models allow evaluation and estimation of the execution time of any given iterative tomographic reconstruction workflow that runs on geographically distributed resources. For the latter challenge, a workflow management system is built, which can automate the execution of workflows and minimize the user interaction with the underlying infrastructure. The system utilizes Globus to perform secure and efficient data transfer operations. The proposed models and the workflow management system are evaluated by using three high-performance computing and two storage resources, all of which are geographically distributed. Workflows were created with different computational requirements using two compute-intensive tomographic reconstruction algorithms. Experimental evaluation shows that the proposed models and system can be used for selecting the optimum resources, which in turn can provide up to 3.13× speedup (on experimented resources). Moreover, the error rates of the models range between 2.1 and 23.3% (considering workflow execution times), where the accuracy of the model estimations increases with higher computational demands in reconstruction tasks.
منابع مشابه
Dynamic configuration and collaborative scheduling in supply chains based on scalable multi-agent architecture
Due to diversified and frequently changing demands from customers, technological advances and global competition, manufacturers rely on collaboration with their business partners to share costs, risks and expertise. How to take advantage of advancement of technologies to effectively support operations and create competitive advantage is critical for manufacturers to survive. To respond to these...
متن کاملCrossroads - Spring 2010
14 Spring 2010/ Vol. 16, No. 3 www.acm.org/crossroads Crossroads Besides public data repositories, scientific collaborations maintain community-wide data resources. For example, in gravitational-wave physics, the Laser Interferometer Gravitational-Wave Observatory [3] maintains geographically distributed repositories holding time-series data collected by the instruments and their associated met...
متن کاملDemo: eQoSystem – Supporting Fluid Distributed Service-Oriented Workflows
Many distributed applications have emerged asWebmashups [1], as well as loosely-coupled decentralized services predominant in a business ecosystem [3]. Many of these applications are implemented as service-oriented workflows and operated over cloud infrastructures. As a result, these applications demand agile development processes and low-touch maintenance life-cycles. Furthermore, in the cloud...
متن کاملDynamic computational workflows: Discovery, optimisation and scheduling
The Grid computing community is converging on a service-oriented architecture in which applications are composed from geographically-distributed, interacting web services, and are expressed in a workflow description language, usually based on XML. Such workflows are viewed as offering a useful representation of service-based applications or applications composed of standalone components that ar...
متن کاملComplexity Analysis and Performance Optimization of Distributed Computing Workflows: From Theory to Practice
The advance of supercomputing technology is expediting the transition in various basic and applied sciences from traditional laboratory-controlled experimental methodologies to modern computational paradigms involving complex numerical model analyses and extreme-scale simulations. These computationbased simulations and analyses have become an essential research and discovery tool in next-genera...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Journal of synchrotron radiation
دوره 23 Pt 4 شماره
صفحات -
تاریخ انتشار 2016